real plant
Real-DRL: Teach and Learn in Reality
Mao, Yanbing, Cai, Yihao, Sha, Lui
This paper introduces the Real-DRL framework for safety-critical autonomous systems, enabling runtime learning of a deep reinforcement learning (DRL) agent to develop safe and high-performance action policies in real plants (i.e., real physical systems to be controlled), while prioritizing safety! The Real-DRL consists of three interactive components: a DRL-Student, a PHY-Teacher, and a Trigger. The DRL-Student is a DRL agent that innovates in the dual self-learning and teaching-to-learn paradigm and the real-time safety-informed batch sampling. On the other hand, PHY-Teacher is a physics-model-based design of action policies that focuses solely on safety-critical functions. PHY-Teacher is novel in its real-time patch for two key missions: i) fostering the teaching-to-learn paradigm for DRL-Student and ii) backing up the safety of real plants. The Trigger manages the interaction between the DRL-Student and the PHY-Teacher. Powered by the three interactive components, the Real-DRL can effectively address safety challenges that arise from the unknown unknowns and the Sim2Real gap. Additionally, Real-DRL notably features i) assured safety, ii) automatic hierarchy learning (i.e., safety-first learning and then high-performance learning), and iii) safety-informed batch sampling to address the learning experience imbalance caused by corner cases. Experiments with a real quadruped robot, a quadruped robot in NVIDIA Isaac Gym, and a cart-pole system, along with comparisons and ablation studies, demonstrate the Real-DRL's effectiveness and unique features.
- North America > United States > Illinois > Champaign County > Urbana (0.14)
- North America > United States > Michigan > Wayne County > Detroit (0.04)
- North America > United States > California > San Francisco County > San Francisco (0.04)
- Europe > Germany > Bavaria > Upper Bavaria > Munich (0.04)
- Information Technology (1.00)
- Transportation > Ground > Road (0.93)
- Information Technology > Artificial Intelligence > Representation & Reasoning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks (1.00)
- Information Technology > Artificial Intelligence > Robots > Locomotion (0.87)
- Information Technology > Artificial Intelligence > Robots > Autonomous Vehicles (0.68)
Simplex-enabled Safe Continual Learning Machine
Cai, Yihao, Cao, Hongpeng, Mao, Yanbing, Sha, Lui, Caccamo, Marco
This paper proposes the SeC-Learning Machine: Simplex-enabled safe continual learning for safety-critical autonomous systems. The SeC-learning machine is built on Simplex logic (that is, ``using simplicity to control complexity'') and physics-regulated deep reinforcement learning (Phy-DRL). The SeC-learning machine thus constitutes HP (high performance)-Student, HA (high assurance)-Teacher, and Coordinator. Specifically, the HP-Student is a pre-trained high-performance but not fully verified Phy-DRL, continuing to learn in a real plant to tune the action policy to be safe. In contrast, the HA-Teacher is a mission-reduced, physics-model-based, and verified design. As a complementary, HA-Teacher has two missions: backing up safety and correcting unsafe learning. The Coordinator triggers the interaction and the switch between HP-Student and HA-Teacher. Powered by the three interactive components, the SeC-learning machine can i) assure lifetime safety (i.e., safety guarantee in any continual-learning stage, regardless of HP-Student's success or convergence), ii) address the Sim2Real gap, and iii) learn to tolerate unknown unknowns in real plants. The experiments on a cart-pole system and a real quadruped robot demonstrate the distinguished features of the SeC-learning machine, compared with continual learning built on state-of-the-art safe DRL frameworks with approaches to addressing the Sim2Real gap.
- Europe > Germany > Bavaria > Upper Bavaria > Munich (0.04)
- North America > United States > Illinois > Champaign County > Urbana (0.04)
- North America > United States > California > Santa Clara County > Palo Alto (0.04)
- Transportation (0.68)
- Education (0.67)
- Information Technology (0.46)
- Energy (0.46)
Importance of realism in procedurally-generated synthetic images for deep learning: case studies in maize and canola
Khan, Nazifa Azam, Cieslak, Mikolaj, McQuillan, Ian
Artificial neural networks are often used to identify features of crop plants. However, training their models requires many annotated images, which can be expensive and time-consuming to acquire. Procedural models of plants, such as those developed with Lindenmayer-systems (L-systems) can be created to produce visually realistic simulations, and hence images of plant simulations, where annotations are implicitly known. These synthetic images can either augment or completely replace real images in training neural networks for phenotyping tasks. In this paper, we systematically vary amounts of real and synthetic images used for training in both maize and canola to better understand situations where synthetic images generated from L-systems can help prediction on real images. This work also explores the degree to which realism in the synthetic images improves prediction. We have five different variants of a procedural canola model (these variants were created by tuning the realism while using calibration), and the deep learning results showed how drastically these results improve as the canola synthetic images are made to be more realistic. Furthermore, we see how neural network predictions can be used to help calibrate L-systems themselves, creating a feedback loop.
- North America > Canada > Alberta > Census Division No. 6 > Calgary Metropolitan Region > Calgary (0.14)
- Asia > Middle East > Jordan (0.04)
- North America > United States > Nebraska > Lancaster County > Lincoln (0.04)
- (3 more...)
Physical Deep Reinforcement Learning: Safety and Unknown Unknowns
Cao, Hongpeng, Mao, Yanbing, Sha, Lui, Caccamo, Marco
In this paper, we propose the Phy-DRL: a physics-model-regulated deep reinforcement learning framework for safety-critical autonomous systems. The Phy-DRL is unique in three innovations: i) proactive unknown-unknowns training, ii) conjunctive residual control (i.e., integration of data-driven control and physics-model-based control) and safety- \& stability-sensitive reward, and iii) physics-model-based neural network editing, including link editing and activation editing. Thanks to the concurrent designs, the Phy-DRL is able to 1) tolerate unknown-unknowns disturbances, 2) guarantee mathematically provable safety and stability, and 3) strictly comply with physical knowledge pertaining to Bellman equation and reward. The effectiveness of the Phy-DRL is finally validated by an inverted pendulum and a quadruped robot. The experimental results demonstrate that compared with purely data-driven DRL, Phy-DRL features remarkably fewer learning parameters, accelerated training and enlarged reward, while offering enhanced model robustness and safety assurance.
- North America > United States (0.46)
- Europe (0.28)
Physical Deep Reinforcement Learning Towards Safety Guarantee
Cao, Hongpeng, Mao, Yanbing, Sha, Lui, Caccamo, Marco
Deep reinforcement learning (DRL) has achieved tremendous success in many complex decision-making tasks of autonomous systems with high-dimensional state and/or action spaces. However, the safety and stability still remain major concerns that hinder the applications of DRL to safety-critical autonomous systems. To address the concerns, we proposed the Phy-DRL: a physical deep reinforcement learning framework. The Phy-DRL is novel in two architectural designs: i) Lyapunov-like reward, and ii) residual control (i.e., integration of physics-model-based control and data-driven control). The concurrent physical reward and residual control empower the Phy-DRL the (mathematically) provable safety and stability guarantees. Through experiments on the inverted pendulum, we show that the Phy-DRL features guaranteed safety and stability and enhanced robustness, while offering remarkably accelerated training and enlarged reward.
- Europe > Germany > Bavaria > Upper Bavaria > Munich (0.05)
- North America > United States > Michigan > Wayne County > Detroit (0.04)
- North America > United States > Illinois > Champaign County > Urbana (0.04)
- (2 more...)
Robot with origami leaves can follow the sun like a real plant
Many plants naturally bend towards bright light. Now a robot has been built that copies a technique plants use to do the same thing. Creating a robot that can sense and adjust automatically to its environment without any need for programming or maintenance is one of the major goals of robotics. A machine that could control and regulate itself in this way can then behave like a living organism, says Bilge Baytekin at Bilkent University in Turkey.
Indoor smart garden uses artificial intelligence to grow real plants
Urban agriculture is a booming trend these days, and now, even the most agriculturally challenged of us can get in on the fun with this AI garden. No one will know if you lack a green thumb so long as you have a smart garden. No matter how agriculturally dumb you may be, artificial intelligence is here to help. It comes in the form of AVA Byte from AVA Technologies. Heralded as the first indoor smart garden to combine AI technology, including machine learning, and a straightforward user experience, you can have a farmer's market right in your own home.